Apple Intelligence
Apple Intelligence is the personal intelligence system that puts powerful generative models right at the core of your iPhone, iPad, and Mac. It powers incredible new features to help people communicate, work, and express themselves.
Writing Tools are available system-wide, and help users rewrite, proofread, and summarize text. The Image Playground API delivers an easy-to-use experience where people can create fun, playful images right in your app. Genmoji bring fun new ways to communicate by providing the ability to create an emoji for any occasion. And Siri will be able to access text displayed in your app, and take hundreds of new actions in and across apps using the App Intents you make available.
Bring your advanced games to Mac
Accelerate your timeline, whether you’re bringing an existing game or one that’s currently in development. With the Game Porting Toolkit 2, it’s never been easier to bring advanced games to Mac, iPad, and iPhone — and to deliver great experiences for players.
Metal
Metal powers hardware-accelerated graphics on Apple platforms. Now you can naturally port apps from other platforms with added support for global bindings and extensions to ray tracing. Residency sets let you quickly and easily group resources to manage residency. And it’s even simpler to target Mac, iPad, and iPhone with the same code, thanks to a unified API and shader build process.
App Intents
New orchestration capabilities provided by Apple Intelligence and significant enhancements to App Intents will enable Siri to take hundreds of new actions in and across apps. Using the Transferable API, File Representations, Item Providers, Spotlight Index, and more, you can make your entities more meaningful to the platform. Plus, you can provide powerful functionality in Siri and Spotlight. You can also explore new APIs for error handling, deferred properties, and associated enums.
Machine learning
Core ML
Updates to Core ML will help you optimize and run advanced generative machine learning and AI models on device faster and more efficiently. Core ML Tools offer more granular and composable weight compression techniques to help you bring your large language models and diffusion models to Apple silicon. Models can now hold multiple functions and efficiently manage state, enabling more flexible and efficient execution of large language models and adapters. The Core ML framework also adds a new MLTensor type that provides an efficient, simple, and familiar API for expressing operations on multi-dimensional arrays. And Core ML performance reports in Xcode have been updated to provide more insight into support and estimated cost of each operation in your model.
Create ML
Object tracking, the first spatial computing template, is designed to help you track real world objects in your visionOS app. Enhance your customized model training workflow with the new data preview functionality in the Create ML app and new Swift APIs from Create ML Components that help you create time series models directly within your app.
Machine learning APIs
The new Translation framework allows you to translate text across different languages in your app. The Vision framework API has been redesigned to leverage modern Swift features, and also supports two new features: image aesthetics and holistic body pose. And the Natural Language framework offers extended language support with multilingual contextual embedding.
RealityKit 4
RealityKit 4 aligns its rich feature set across iPhone, iPad, Mac, and Apple Vision Pro. Reality Composer Pro, a new tool that launched with Apple Vision Pro, enables development of spatial apps on all these platforms.
Shaders built with MaterialX, portals, particles, and many other features can now be used with RealityView on all four platforms. This includes APIs for adding materials, shader-based hover effects, and virtual lighting, as well as new features — like blend shapes, inverse kinematics, skeletal poses, and animation timelines — that expand character animation capabilities.
RealityKit 4 also provides more direct access to rendering with new APIs for low-level mesh and textures, which work with Metal compute shaders. And because Xcode view debugging now supports inspecting 3D scene content, it’s easier than ever to inspect and debug your RealityKit content.
Passkeys
Passkeys are a replacement for passwords that are more secure, easier to use, and can’t be phished. They offer faster sign-in, fewer password resets, and reduced support costs. Use the new automatic passkey upgrade API to create a passkey when someone signs in to your app and let them know that the passkey was saved — all without interrupting their flow.
App Store and StoreKit
Find out how to nominate your apps for featuring on the App Store and share exciting moments (like a version launch) with marketing assets generated for you. Enhancements to StoreKit views give you more flexibility and customization options when building your merchandising experiences. Improvements to StoreKit testing in Xcode and the Apple Sandbox environment help you test additional purchase scenarios. Offer codes for macOS apps let you provide discounted or free subscriptions for a period of time. And win-back offers give you a new way to re-engage previous subscribers.